Lower bounds on minimax rates for nonparametric regression with additive sparsity and smoothness
نویسندگان
چکیده
We study minimax rates for estimating high-dimensional nonparametric regression models with sparse additive structure and smoothness constraints. More precisely, our goal is to estimate a function f∗ : R → R that has an additive decomposition of the form f(X1, . . . ,Xp) = ∑ j∈S h ∗ j (Xj), where each component function h ∗ j lies in some class H of “smooth” functions, and S ⊂ {1, . . . , p} is an unknown subset with cardinality s = |S|. Given n i.i.d. observations of f∗(X) corrupted with additive white Gaussian noise where the covariate vectors (X1,X2,X3, ...,Xp) are drawn with i.i.d. components from some distribution P, we determine lower bounds on the minimax rate for estimating the regression function with respect to squared-L(P) error. Our main result is a lower bound on the minimax rate that scales as max ( s log(p/s) n , s ǫ 2 n(H) ) . The first term reflects the sample size required for performing subset selection, and is independent of the function class H. The second term s ǫn(H) is an s-dimensional estimation term corresponding to the sample size required for estimating a sum of s univariate functions, each chosen from the function class H. It depends linearly on the sparsity index s but is independent of the global dimension p. As a special case, if H corresponds to functions that are m-times differentiable (an m-order Sobolev space), then the s-dimensional estimation term takes the form sǫn(H) ≍ s n−2m/(2m+1). Either of the two terms may be dominant in different regimes, depending on the relation between the sparsity and smoothness of the additive decomposition.
منابع مشابه
Minimax-optimal rates for high-dimensional sparse additive models over kernel classes
Sparse additive models are families of d-variate functions that have the additive decomposition f = ∑ j∈S f ∗ j , where S is an unknown subset of cardinality s ≪ d. We consider the case where each component function f j lies in a reproducing kernel Hilbert space, and analyze an l1 kernel-based method for estimating the unknown function f . Working within a highdimensional framework that allows ...
متن کاملMinimax-Optimal Rates For Sparse Additive Models Over Kernel Classes Via Convex Programming
Sparse additive models are families of d-variate functions that have the additive decomposition f∗ = ∑ j∈S f ∗ j , where S is a unknown subset of cardinality s d. We consider the case where each component function f∗ j lies in a reproducing kernel Hilbert space, and analyze a simple kernel-based convex program for estimating the unknown function f∗. Working within a high-dimensional framework t...
متن کاملMultivariate Dyadic Regression Trees for Sparse Learning Problems
We propose a new nonparametric learning method based on multivariate dyadic regression trees (MDRTs). Unlike traditional dyadic decision trees (DDTs) or classification and regression trees (CARTs), MDRTs are constructed using penalized empirical risk minimization with a novel sparsity-inducing penalty. Theoretically, we show that MDRTs can simultaneously adapt to the unknown sparsity and smooth...
متن کاملSparse additive regression on a regular lattice
We consider estimation in a sparse additive regression model with the design points on a regular lattice. We establish the minimax convergence rates over Sobolev classes and propose a Fourier-based rate optimal estimator which is adaptive to the unknown sparsity and smoothness of the response function. The estimator is derived within a Bayesian formalism but can be naturally viewed as a penaliz...
متن کاملMinimax Lower Bounds for Noisy Matrix Completion Under Sparse Factor Models
This paper examines fundamental error characteristics for a general class of matrix completion problems, where matrix of interest is a product of two a priori unknown matrices, one of which is sparse, and the observations are noisy. Our main contributions come in the form of minimax lower bounds for the expected per-element squared error for these problems under several noise/corruption models;...
متن کامل